perm filename CMPARE[0,BGB]4 blob sn#109791 filedate 1974-07-10 generic text, type C, neo UTF8
COMMENT ⊗   VALID 00004 PAGES
C REC  PAGE   DESCRIPTION
C00001 00001
C00002 00002	{[C<NαIMAGE COMPARING.λ30P80I425,0JCFA} SECTION 8.
C00005 00003		The compare step of  CRE,  CMPARE, connects the  polygons and
C00010 00004	[COMPARING.]
C00014 ENDMK
C⊗;
{[C;<N;αIMAGE COMPARING.;λ30;P80;I425,0;JCFA} SECTION 8.
{JCFD}    IMAGE COMPARING.
{λ10;W250;JAFA}
	8.0	Introduction to Image Comparing.

{λ30;W0;I950,0;JUFA}
[7.0	Introduction to Image Comparing.]
{JUFA}

	The compare process  is the keystone for completing  the arch
of  a visual feedback  system;  however the compare  need not require
much study if the entities  to be mated can be highly  individualized
so that there is either a monogamous match or nothing. (Although, the
study  of  sophisticated  comparing is  important;  in  a descriptive
approach to vision, when a simple compare doesn't  work, one censures
the object representation rather that the match processor.)

	Three  important  compares  can  be characterized  as  verify
compare, reveal compare  and recognize compare.   The verify  compare
involves finding the corresponding entities between a predicted image
and a perceived  image for the sake of camera calibration and for the
sake  of eliminating  know  landmark  features  from  the  revelation
process.    The reveal  compare  involves  finding the  corresponding
entities in two percieved images,  so that the location and extent of
new landmark  objects  can  be solved.    Finally,   the  recognition
compare involves mating a percieved  entity with one of a set of know
model entities; recognition is done both in the 2-D image domain  and
in the domain of  3-D object models, although the  later will receive
more attention.

	Two  problems of description comparing  are normalization and
segmentation.     Normalization   involves   eliminating   irrelevant
differences such as location,  orientation and lighting. Segmentation
involves  subdividing a complex object into  pieces, so that only the
simple small pieces need be matched.

	The compare step of  CRE,  CMPARE, connects the  polygons and
arcs  of the current  image with  corresponding polygons and  arcs of
the  previous image.    CMPARE  solves  the  problem  of  correlating
features  between  two  similar  images  and  is  composed  four  sub
sections: 

	1. make shape nodes for polygons.
	2. compare and connect polygons one to one.
	3. compare and connect polygons two to one.
	4. compare and connect vertices of connected polygons.

	First,  the shape  nodes of all the polygons of  an image are
computed. The  shape node contains the center  of mass and the lamina
inertia tensor of a polygon.  The lamina inertia tensor of  a polygon
with  N sides  is  computed  by summation  over  N  trapezoids.   The
trapezoid   corresponding  to   each  side  is   formed  by  dropping
perpendiculars  "up"  to the  top  of  the  image  frame;  each  such
trapezoid  consists of  a rectangle  an a  right triangle;  since the
sides of polygons  are directed  vectors the areas  of the  triangles
and rectangles can be  arranged to take positive and  negative values
such  that  a summation  will  describe the  interior  region  of the
polygon  as positive.  The  equations  necessary  for  computing  the
lamina inertia  tensor of a polygon  are collected in a  table in the
postscripts  to  this paper  and  were derived  by  using Goldstein's
Classical Mechanics (1) as a  reference.  The meaning of  the inertia
tensor  is that it  characterizes each  polygon by  a rectangle  of a
certain length and width at a particular location and oriention;  and
of  further  importance  such  inertia  tensors  can  be  added  to
characterize two  or more polygons by a single  rectangle.  It is the
lamina inertia tensor rectangles that are actually compared by CRE.

	Second, all the shapes  of the polygons  of one level of  the
first image are  compared with all the shapes of  the polygons of the
corresponding level  of the second image for nearly exact match.  The
potentially (M*N/2) compares is  avoided by sorting on the  center of
mass locations.   In CRE,  which is intended  for comparing sequences
of pictures of natural scenes; match  for center of mass location  is
tested first  and  most strictly,   followed  by  match for  inertia.
Pointers  between  matching  polygons are  placed  in  the time  link
positions of the polygon nodes and the polygons are considered  to be
mated in time.

[COMPARING.]

	Third,  all  the unmated polygons  of a level  are considered
two  at a time and  a fusion shape node  for each pair is  made.  The
potentially (N*N/2-N) fusion  shapes are avoided  because there is  a
maximum possible  unmated inertia  in the other  image; lo,  if there
are  no unmated polygons in one image  then the extra polygons of the
first image  can be  ignored. In the  event where  there are  unmated
polygons  in  corresponding levels  of  the  two  images, the  fusion
shapes of one  are compared  with the  polygon shapes  of the  other.
The  fusion  (fission)  compare  solves  the  rather  nasty  problem,
illustrated in  figures 9A and 9B of  linking two contour polygons of
one image with a single contour polygon in the next image.

	Fourth, the vertices of  polygons mated in time are  compared
and mated.  To start a  vertex compare,  the vertices of  one polygon
are  translated,   rotated and dilated  to get  that polygon's lamina
inertia tensor  coincidant with  its mate  (or mates).  Conceptually,
each vertex of one polygon  is compared with each vertex of the other
polygon(s)  and  the  mutually  closest  vertices  (closer  than   an
epsilon) are  considered to  be mated.  Actually the potential  (N*M)
compares  is avoided by  a window  splitting scheme similiar  to that
used in hidden line elimination algorithms (like Warnock's).

	The results  of vertex compare  and mate  are illustrated  in
figures 9A and 9D; the compare  execution takes less than a second on
images  such as the  pump,  blocks,  and dolls that  have appeared in
this  paper. The  applications  of  this compare  might  include  the
aiming   of  a  pixel   correlation  comparator  (such   as  Quam's);
recognition and location of an  expected object; or the location  and
extent of  an unknown  object.   It is  this latter application  that
will be described in my forthcoming thesis.